Distance metric learning for graph structured data

نویسندگان

چکیده

Graphs are versatile tools for representing structured data. As a result, variety of machine learning methods have been studied graph data analysis. Although many such depend on the measurement differences between input graphs, defining an appropriate distance metric graphs remains controversial issue. Hence, we propose supervised method classification problem. Our method, named interpretable (IGML), learns discriminative metrics in subgraph-based feature space, which has strong representation capability. By introducing sparsity-inducing penalty weight each subgraph, IGML can identify small number important subgraphs that provide insight into given task. Because our formulation large optimization variables, efficient algorithm uses pruning techniques based safe screening and working set selection is also proposed. An property solution optimality guaranteed because problem formulated as convex strategies only discard unnecessary subgraphs. Furthermore, show applicable to other itemset sequence data, it incorporate vertex-label similarity by using transportation-based subgraph feature. We empirically evaluate computational efficiency performance several benchmark datasets some illustrative examples how identifies from dataset.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Deep Distance Metric Learning with Data Summarization

We present Deep Stochastic Neighbor Compression (DSNC), a framework to compress training data for instance-based methods (such as k-nearest neighbors). We accomplish this by inferring a smaller set of pseudo-inputs in a new feature space learned by a deep neural network. Our framework can equivalently be seen as jointly learning a nonlinear distance metric (induced by the deep feature space) an...

متن کامل

Learning a Mahalanobis distance metric for data clustering and classification

Article history: Received 7 October 2007 Received in revised form 27 February 2008 Accepted 16 May 2008

متن کامل

Bayesian Distance Metric Learning

This thesis explores the use of Bayesian distance metric learning (Bayes-dml) for the task of speaker verification using the i-vector feature representation. We propose a framework that explores the distance constraints between i-vector pairs from the same speaker and different speakers. With an approximation of the distance metric as a weighted covariance matrix of the top eigenvectors from th...

متن کامل

Distance Metric Learning Revisited

The success of many machine learning algorithms (e.g. the nearest neighborhood classification and k-means clustering) depends on the representation of the data as elements in a metric space. Learning an appropriate distance metric from data is usually superior to the default Euclidean distance. In this paper, we revisit the original model proposed by Xing et al. [24] and propose a general formu...

متن کامل

Hamming Distance Metric Learning

Motivated by large-scale multimedia applications we propose to learn mappings from high-dimensional data to binary codes that preserve semantic similarity. Binary codes are well suited to large-scale applications as they are storage efficient and permit exact sub-linear kNN search. The framework is applicable to broad families of mappings, and uses a flexible form of triplet ranking loss. We ov...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Machine Learning

سال: 2021

ISSN: ['0885-6125', '1573-0565']

DOI: https://doi.org/10.1007/s10994-021-06009-3